Confidence-rated Regression by Localized Median Boosting
نویسنده
چکیده
In this paper we describe and analyze LOCMEDBOOST, an algorithm that boosts regressors with input dependent weights. The algorithm is a synthesis of median boosting [1] and localized boosting [2, 3, 4], and unifies the advantages of the two approaches. We prove boostingtype convergence of the algorithm and give clear conditions for the convergence of the robust training error, where robustness is measured in terms of the expert population and with respect to the underlying confidence estimate. We extend Rätsch and Warmuth’s results [5] on efficient margin maximization to show that the algorithm can converge to maximum achievable margin in a finite number of steps. We also extend probabilistic bounds on the generalization error derived for ADABOOST. The results provide bounds on the confidence-interval-type error and qualitatively justify the algorithmic objective of the minimization of the robust error. Finally, we present some promising experimental results on synthetic and benchmark data sets.
منابع مشابه
Robust Regression by Boosting the Median
Most boosting regression algorithms use the weighted average of base regressors as their final regressor. In this paper we analyze the choice of the weighted median. We propose a general boosting algorithm based on this approach. We prove boosting-type convergence of the algorithm and give clear conditions for the convergence of the robust training error. The algorithm recovers ADABOOST and ADA...
متن کاملGeneralization Error and Algorithmic Convergence of Median Boosting
We have recently proposed an extension of ADABOOST to regression that uses the median of the base regressors as the final regressor. In this paper we extend theoretical results obtained for ADABOOST to median boosting and to its localized variant. First, we extend recent results on efficient margin maximizing to show that the algorithm can converge to the maximum achievable margin within a pres...
متن کاملImproving Algorithms for Boosting
Motivated by results in information-theory, we describe a modification of the popular boosting algorithm AdaBoost and assess its performance both theoretically and empirically. We provide theoretical and empirical evidence that the proposed boosting scheme will have lower training and testing error than the original (nonconfidence-rated) version of AdaBoost. Our modified boosting algorithm and ...
متن کاملIncome gaps in self-rated poor health and its association with life expectancy in 245 districts of Korea
OBJECTIVES To examine the income gaps associated with self-rated poor health at the district level in Korea and to identify the geographical correlations between self-rated poor health, life expectancy, and the associated income gaps. METHODS We analyzed data for 1,578,189 participants from the Community Health Survey of Korea collected between 2008 and 2014. The age-standardized prevalence o...
متن کاملOutlier Detection by Boosting Regression Trees
A procedure for detecting outliers in regression problems is proposed. It is based on information provided by boosting regression trees. The key idea is to select the most frequently resampled observation along the boosting iterations and reiterate after removing it. The selection criterion is based on Tchebychev’s inequality applied to the maximum over the boosting iterations of ...
متن کامل